National Repository of Grey Literature 51 records found  1 - 10nextend  jump to record: Search took 0.00 seconds. 
Query Answering over Wikipedia for Mobile Devices on the Android Platform
Kováč, Andrej ; Otrusina, Lubomír (referee) ; Smrž, Pavel (advisor)
p { margin-bottom: 0.1in; direction: ltr; line-height: 120%; text-align: left; widows: 2; orphans: 2; }p.western { font-family: "Times New Roman",serif; }p.cjk { font-family: "Times New Roman"; }p.ctl { font-family: "Times New Roman"; font-size: 12pt; }a:link { color: rgb(0, 0, 255); } This bachelor thesis deals with the development of a system for query answering over Wikipedia for mobile devices running Android operating system. In this technical report theoretical knowledge related to this topic is described as well as the implementation process of a server system and client side application. Part of this thesis is dedicated to testing of the system and in the final part the potential for future development is drafted.
Generating Code from Textual Description of Functionality
Kačur, Ján ; Ondřej, Karel (referee) ; Smrž, Pavel (advisor)
The aim of this thesis was to design and implement system for code generation from textual description of functionality. In total, 2 systems were implemented. One of them served its purpose as a control prototype, the second one was the main product of this thesis. I focused on using smaller non-pre-trained models. Both systems used Transformer type model as their cores. The second system, unlike the first, used syntactic decomposition of both code and textual descriptions. Data used in both systems originated from project CodeSearchNet. Targer programming language to generate was Python. The second system achieved better quantitative results than the first one, with accuracy of 85% versus 60%. The system managed to auto-complete correct code to finish the function definition, with bigger time delay. This thesis is almost exclusively dedicated to the second system.
Word2vec Models with Added Context Information
Šůstek, Martin ; Rozman, Jaroslav (referee) ; Zbořil, František (advisor)
This thesis is concerned with the explanation of the word2vec models. Even though word2vec was introduced recently (2013), many researchers have already tried to extend, understand or at least use the model because it provides surprisingly rich semantic information. This information is encoded in N-dim vector representation and can be recall by performing some operations over the algebra. As an addition, I suggest a model modifications in order to obtain different word representation. To achieve that, I use public picture datasets. This thesis also includes parts dedicated to word2vec extension based on convolution neural network.
Syntactic Analyzer for Czech Language
Beneš, Vojtěch ; Otrusina, Lubomír (referee) ; Kouřil, Jan (advisor)
Master’s thesis describes theoretical basics, solution design, and implementation of constituency (phrasal) parser for Czech language, which is based on a part of speech association into phrases. Created program works with manually built and annotated Czech sample corpus to generate probabilistic context free grammar within runtime machine learning. Parser implementation, based on extended CKY algorithm, then for the input Czech sentence decides if the sentence can be generated by the created grammar and for the positive cases constructs the most probable derivation tree. This result is then compared with the expected parse to evaluate constituency parser success rate.
Methods of Document Summarization on the Web
Belica, Michal ; Očenášek, Pavel (referee) ; Bartík, Vladimír (advisor)
The work deals with automatic summarization of documents in HTML format. As a language of web documents, Czech language has been chosen. The project is focused on algorithms of text summarization. The work also includes document preprocessing for summarization and conversion of text into representation suitable for summarization algorithms. General text mining is also briefly discussed but the project is mainly focused on the automatic document summarization. Two simple summarization algorithms are introduced. Then, the main attention is paid to an advanced algorithm that uses latent semantic analysis. Result of the work is a design and implementation of summarization module for Python language. Final part of the work contains evaluation of summaries generated by implemented summarization methods and their subjective comparison of the author.
A Classification of a Syndicated Content
Matušov, Izidor ; Očenášek, Pavel (referee) ; Smrčka, Aleš (advisor)
This work deals with a classification of a syndicated content as the possible way of organizing the content. The classification uses algorithms for natural language processing. The main contribution is applying word sense disambiguation algorithm for enhancing the classification, eliminating the learning stage, and using a readability test for improving user experience. The application is implemented as an extensible server-client model. The future work is discussed in the end.
XML Databases for Dictionary Data Management
Samia, Michel ; Dytrych, Jaroslav (referee) ; Smrž, Pavel (advisor)
The following diploma thesis deals with dictionary data processing, especially those in XML based formats. At first, the reader is acquainted with linguistic and lexicographical terms used in this work. Then particular lexicographical data format types and specific formats are introduced. Their advantages and disadvantages are discussed as well. According to previously set criteria, the LMF format has been chosen for design and implementation of Python application, which focuses especially on intelligent merging of more dictionaries into one. After passing all unit tests, this application has been used for processing LMF dictionaries, located on the faculty server of the research group for natural language processing. Finally, the advantages and disadvantages of this application are discussed and ways of further usage and extension are suggested.
Methods of Information Extraction
Adamček, Adam ; Smrž, Pavel (referee) ; Kouřil, Jan (advisor)
The goal of information extraction is to retrieve relational data from texts written in natural human language. Applications of such obtained information is wide - from text summarization, through ontology creation up to answering questions by QA systems. This work describes design and implementation of a system working in computer cluster which transforms a dump of Wikipedia articles to a set of extracted information that is stored in distributed RDF database with a possibility to query it using created user interface.
Machine Learning for Natural Language Question Answering
Sasín, Jonáš ; Fajčík, Martin (referee) ; Smrž, Pavel (advisor)
This thesis deals with natural language question answering using Czech Wikipedia. Question answering systems are experiencing growing popularity, but most of them are developed for English. The main purpose of this work is to explore possibilities and datasets available and create such system for Czech. In the thesis I focused on two approaches. One of them uses English model ALBERT and machine translation of passages. The other one utilizes the multilingual BERT. Several variants of the system are compared in this work. Possibilities of relevant passage retrieval are also discussed. Standard evaluation is provided for every variant of the tested system. The best system version has been evaluated on the SQAD v3.0 dataset, reaching 0.44 EM and 0.55 F1 score, which is an excellent result compared to other existing systems. The main contribution of this work is the analysis of existing possibilities and setting a benchmark for further development of better systems for Czech.
Chatbot for Smart Cities
Jusko, Ján ; Herout, Adam (referee) ; Zemčík, Pavel (advisor)
The aim of this work is to simplify access to information for citizens of the city of Brno and at the same time to innovate the way of communication between the citizen and his city. The problem is solved by creating a conversational agent - chatbot Kroko. Using artificial intelligence and a Czech language analyzer, the agent is able to understand and respond to a certain set of textual, natural language queries. The agent is available on the Messenger platform and has a knowledge base that includes data provided by the city council. After conducting an extensive user testing on a total of 76 citizens of the city, it turned out that up to 97\% of respondents like the idea of a city-oriented chatbot and can imagine using it regularly. The main finding of this work is that the general public can easily adopt and effectively use a chatbot. The results of this work motivate further development of practical applications of conversational agents.

National Repository of Grey Literature : 51 records found   1 - 10nextend  jump to record:
Interested in being notified about new results for this query?
Subscribe to the RSS feed.